Reinforcement Learning and Animat Emotions

نویسنده

  • Ian Wright
چکیده

Emotional states, such as happiness or sadness, pose particular problems for information processing theories of mind. Hedonic components of states, unlike cognitive components, lack representational content. Research within Artiicial Life, in particular the investigation of adaptive agent architectures, provides insights into the dynamic relationship between motivation, the ability of control sub-states to gain access to limited processing resources, and prototype emotional states. Holland's learning classiier system provides a concrete example of this relationship, demonstrating sim-plèemotion-like' states, much as a thermostat demonstrates simplèbelief-like' and`desire-like' states. This leads to the conclusion that valency, a particular form of pleasure or displeasure , is a self-monitored process of credit-assignment. The importance of the movement of a domain-independent representation of utility within adaptive archi-tectures is stressed. Existing information processing theories of emotion can be enriched by a `circulation of value' design hypothesis. Implications for the development of emotional animats are considered.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Dynamic Expectancy: An Approach to Behaviour Shaping Using a New Method of Reinforcement Learning

This paper is concerned with issues relating to the source of reward and reinforcement with potential application to various robot learning and behaviour shaping situations (Dorigo and Colombetti, 1994; Lin, 1991, Maclin and Shavlik, 1996). The conventional approach to behaviour shaping by reinforcement learning is to present “reward” to an animal, animat or robot immediately following the perf...

متن کامل

Incremental Self-improvement for Life-time Multi-agent Reinforcement Learning

Previous approaches to multi-agent reinforcement learning are either very limited or heuris-tic by nature. The main reason is: each agent's or \animat's" environment continually changes because the other learning animats keep changing. Traditional reinforcement learning algorithms cannot properly deal with this. Their convergence theorems require repeatable trials and strong (typically Markovia...

متن کامل

Lookahead And Latent Learning In ZCS

Learning Classifier Systems use reinforcement learning, evolutionary computing and/or heuristics to develop adaptive systems. This paper extends the ZCS Learning Classifier System to improve its internal modelling capabilities. Initially, results are presented which show performance in a traditional reinforcement learning task incorporating lookahead within the rule structure. Then a mechanism ...

متن کامل

Actor-Critic Models of Reinforcement Learning in the Basal Ganglia: From Natural to Artificial Rats

Since 1995, numerous Actor-Critic architectures for reinforcement learning have been proposed as models of dopamine-like reinforcement learning mechanisms in the rat’s basal ganglia. However, these models were usually tested in different tasks, and it is then difficult to compare their efficiency for an autonomous animat. We present here the comparison of four architectures in an animat as it p...

متن کامل

Learning To Do Without Cognitiony

In this paper we show that a phenomenon in animal learning theory (the outcome devaluation eeect) for which there is dispute over whether explicit representations and symbolic reasoning is required for its performance , does not require such things. This is done using a reactive motivational model, previously inspired from ethological thought, to which some simple reinforcement learning rules a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1996